Goto

Collaborating Authors

 calibration module


b3b43aeeacb258365cc69cdaf42a68af-Paper.pdf

Neural Information Processing Systems

We present an approach for lifelong/continual learning of convolutional neural networks (CNN) that does not suffer from the problem of catastrophic forgetting when moving from onetask totheother. Weshowthat theactivation maps generated by the CNN trained on the old task can be calibrated using very few calibration parameters, to become relevant to the new task.


Calibrating CNNs for Lifelong Learning

Neural Information Processing Systems

We present an approach for lifelong/continual learning of convolutional neural networks (CNN) that does not suffer from the problem of catastrophic forgetting when moving from one task to the other. We show that the activation maps generated by the CNN trained on the old task can be calibrated using very few calibration parameters, to become relevant to the new task.


Accurate Parameter-Efficient Test-Time Adaptation for Time Series Forecasting

Medeiros, Heitor R., Sharifi-Noghabi, Hossein, Oliveira, Gabriel L., Irandoust, Saghar

arXiv.org Artificial Intelligence

Real-world time series often exhibit a non-stationary nature, degrading the performance of pre-trained forecasting models. Test-Time Adaptation (TTA) addresses this by adjusting models during inference, but existing methods typically update the full model, increasing memory and compute costs. We propose PETSA, a parameter-efficient method that adapts forecasters at test time by only updating small calibration modules on the input and output. PETSA uses low-rank adapters and dynamic gating to adjust representations without retraining. To maintain accuracy despite limited adaptation capacity, we introduce a specialized loss combining three components: (1) a robust term, (2) a frequency-domain term to preserve periodicity, and (3) a patch-wise structural term for structural alignment. PETSA improves the adaptability of various forecasting backbones while requiring fewer parameters than baselines. Experimental results on benchmark datasets show that PETSA achieves competitive or better performance across all horizons. Our code is available at: https://github.com/BorealisAI/PETSA


Review for NeurIPS paper: Calibrating CNNs for Lifelong Learning

Neural Information Processing Systems

Summary and Contributions: Update: My initial review noted two main issues with the paper: reliance on the initial model, and the use of task labels during the test phase. The author response addresses the first question, but misses the point on the second one. And this alone is not sufficient to strongly influence my overall rating. In my understanding, several previous methods, such as LwF, iCaRL highlighted in the author response, classify samples without the knowledge of which group of classes (i.e., old or new) they belong to. In other words, they only use a single framework that can identify samples from any of the old or the new classes, without additional information.


Review for NeurIPS paper: Calibrating CNNs for Lifelong Learning

Neural Information Processing Systems

The paper proposes a continual learning approach for CNN models. This is achieved through spatial and channel-wise calibration modules, one for each new task. These calibration modules are introduced between each pair of consecutive layers in the original base model. The base model is learnt on the first task, and training data from the subsequent tasks is used to learn the calibration modules. Extensive experiments show the superiority of the proposed method in terms of accuracies, with minimal computation and storage overhead. It is important to emphasize that the proposed approach requires task labels in the test phase.


A Self-boosted Framework for Calibrated Ranking

Zhang, Shunyu, Liu, Hu, Bao, Wentian, Yu, Enyun, Song, Yang

arXiv.org Artificial Intelligence

Scale-calibrated ranking systems are ubiquitous in real-world applications nowadays, which pursue accurate ranking quality and calibrated probabilistic predictions simultaneously. For instance, in the advertising ranking system, the predicted click-through rate (CTR) is utilized for ranking and required to be calibrated for the downstream cost-per-click ads bidding. Recently, multi-objective based methods have been wildly adopted as a standard approach for Calibrated Ranking, which incorporates the combination of two loss functions: a pointwise loss that focuses on calibrated absolute values and a ranking loss that emphasizes relative orderings. However, when applied to industrial online applications, existing multi-objective CR approaches still suffer from two crucial limitations. First, previous methods need to aggregate the full candidate list within a single mini-batch to compute the ranking loss. Such aggregation strategy violates extensive data shuffling which has long been proven beneficial for preventing overfitting, and thus degrades the training effectiveness. Second, existing multi-objective methods apply the two inherently conflicting loss functions on a single probabilistic prediction, which results in a sub-optimal trade-off between calibration and ranking. To tackle the two limitations, we propose a Self-Boosted framework for Calibrated Ranking (SBCR).


FishRecGAN: An End to End GAN Based Network for Fisheye Rectification and Calibration

Shen, Xin, Joo, Kyungdon, Oh, Jean

arXiv.org Artificial Intelligence

We propose an end-to-end deep learning approach to rectify fisheye images and simultaneously calibrate camera intrinsic and distortion parameters. Our method consists of two parts: a Quick Image Rectification Module developed with a Pix2Pix GAN and Wasserstein GAN (W-Pix2PixGAN), and a Calibration Module with a CNN architecture. Our Quick Rectification Network performs robust rectification with good resolution, making it suitable for constant calibration in camera-based surveillance equipment. To achieve high-quality calibration, we use the straightened output from the Quick Rectification Module as a guidance-like semantic feature map for the Calibration Module to learn the geometric relationship between the straightened feature and the distorted feature. We train and validate our method with a large synthesized dataset labeled with well-simulated parameters applied to a perspective image dataset. Our solution has achieved robust performance in high-resolution with a significant PSNR value of 22.343.


High Frequency Residual Learning for Multi-Scale Image Classification

Cheng, Bowen, Xiao, Rong, Wang, Jianfeng, Huang, Thomas, Zhang, Lei

arXiv.org Machine Learning

We present a novel high frequency residual learning framework, which leads to a highly efficient multi-scale network (MSNet) architecture for mobile and embedded vision problems. The architecture utilizes two networks: a low resolution network to efficiently approximate low frequency components and a high resolution network to learn high frequency residuals by reusing the upsampled low resolution features. With a classifier calibration module, MSNet can dynamically allocate computation resources during inference to achieve a better speed and accuracy trade-off. We evaluate our methods on the challenging ImageNet-1k dataset and observe consistent improvements over different base networks. On ResNet-18 and MobileNet with alpha=1.0, MSNet gains 1.5% accuracy over both architectures without increasing computations. On the more efficient MobileNet with alpha=0.25, our method gains 3.8% accuracy with the same amount of computations.